Skip to content

[CI] Bump num_speculative_tokens to 3 in nightly DeepSeek tests#35882

Merged
robertgshaw2-redhat merged 2 commits intovllm-project:mainfrom
MatthewBonanni:test_sparse_mtp_3
Mar 3, 2026
Merged

[CI] Bump num_speculative_tokens to 3 in nightly DeepSeek tests#35882
robertgshaw2-redhat merged 2 commits intovllm-project:mainfrom
MatthewBonanni:test_sparse_mtp_3

Conversation

@MatthewBonanni
Copy link
Copy Markdown
Collaborator

@MatthewBonanni MatthewBonanni commented Mar 3, 2026

Purpose

Now that #34552 has landed, both standard and sparse MLA support MTP > 1. We should exercise this in CI.

Test Plan

LM Eval Large Models (H200) should pass CI

Test Result

TBD


Essential Elements of an Effective PR Description Checklist
  • The purpose of the PR, such as "Fix some issue (link existing issues this PR will resolve)".
  • The test plan, such as providing test command.
  • The test results, such as pasting the results comparison before and after, or e2e results
  • (Optional) The necessary documentation update, such as updating supported_models.md and examples for a new model.
  • (Optional) Release notes update. If your change is user facing, please update the release notes draft in the Google Doc.

Signed-off-by: Matthew Bonanni <mbonanni@redhat.com>
@MatthewBonanni MatthewBonanni requested a review from mgoin as a code owner March 3, 2026 16:45
@MatthewBonanni MatthewBonanni added the ready ONLY add when PR is ready to merge/full CI is needed label Mar 3, 2026
@mergify mergify Bot added the deepseek Related to DeepSeek models label Mar 3, 2026
Copy link
Copy Markdown
Contributor

@gemini-code-assist gemini-code-assist Bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

This pull request updates the nightly DeepSeek test configurations by increasing num_speculative_tokens from 1 to 3 for the DeepSeek-V3.2 model. This change applies to both data-parallel and tensor-parallel test setups. The goal is to exercise the recently added support for multi-token prediction (MTP) with more than one token in the CI pipeline, thereby enhancing test coverage for this new feature. The changes are straightforward and align with the stated purpose of the pull request.

Signed-off-by: Matthew Bonanni <mbonanni@redhat.com>
@robertgshaw2-redhat robertgshaw2-redhat enabled auto-merge (squash) March 3, 2026 16:57
@robertgshaw2-redhat robertgshaw2-redhat merged commit 8e1fd5b into vllm-project:main Mar 3, 2026
12 checks passed
cmacboyd added a commit to cmacboyd/vllm that referenced this pull request Mar 3, 2026
…lm-project#35882)

Signed-off-by: Matthew Bonanni <mbonanni@redhat.com>

Signed-off-by: cmacboyd <42761774+cmacboyd@users.noreply.github.com>
Copilot AI pushed a commit to machov/vllm that referenced this pull request Mar 10, 2026
avinashsingh77 pushed a commit to avinashsingh77/vllm that referenced this pull request Mar 12, 2026
wendyliu235 pushed a commit to wendyliu235/vllm-public that referenced this pull request Mar 18, 2026
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

deepseek Related to DeepSeek models ready ONLY add when PR is ready to merge/full CI is needed

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants